11 research outputs found
Towards a unified approach
"Decision-making in the presence of uncertainty is a pervasive computation.
Latent variable decoding—inferring hidden causes underlying visible
effects—is commonly observed in nature, and it is an unsolved challenge
in modern machine learning.
On many occasions, animals need to base their choices on uncertain
evidence; for instance, when deciding whether to approach or avoid an
obfuscated visual stimulus that could be either a prey or a predator. Yet,
their strategies are, in general, poorly understood.
In simple cases, these problems admit an optimal, explicit solution.
However, in more complex real-life scenarios, it is difficult to determine the
best possible behavior. The most common approach in modern machine
learning relies on artificial neural networks—black boxes that map each
input to an output. This input-output mapping depends on a large number
of parameters, the weights of the synaptic connections, which are optimized
during learning.(...)
Beyond topological persistence: Starting from networks
Persistent homology enables fast and computable comparison of topological
objects. However, it is naturally limited to the analysis of topological
spaces. We extend the theory of persistence, by guaranteeing robustness and
computability to significant data types as simple graphs and quivers. We focus
on categorical persistence functions that allow us to study in full generality
strong kinds of connectedness such as clique communities, -vertex and
-edge connectedness directly on simple graphs and monic coherent categories.Comment: arXiv admin note: text overlap with arXiv:1707.0967
Parametric machines: a fresh approach to architecture search
Using tools from category theory, we provide a framework where artificial
neural networks, and their architectures, can be formally described. We first
define the notion of machine in a general categorical context, and show how
simple machines can be combined into more complex ones. We explore finite- and
infinite-depth machines, which generalize neural networks and neural ordinary
differential equations. Borrowing ideas from functional analysis and kernel
methods, we build complete, normed, infinite-dimensional spaces of machines,
and discuss how to find optimal architectures and parameters -- within those
spaces -- to solve a given computational problem. In our numerical experiments,
these kernel-inspired networks can outperform classical neural networks when
the training dataset is small.Comment: 31 pages, 4 figure
Persistence-based operators in machine learning
Artificial neural networks can learn complex, salient data features to
achieve a given task. On the opposite end of the spectrum, mathematically
grounded methods such as topological data analysis allow users to design
analysis pipelines fully aware of data constraints and symmetries. We introduce
a class of persistence-based neural network layers. Persistence-based layers
allow the users to easily inject knowledge about symmetries (equivariance)
respected by the data, are equipped with learnable weights, and can be composed
with state-of-the-art neural architectures
Dependent Optics
A wide variety of bidirectional data accessors, ranging from mixed optics to
functor lenses, can be formalized within a unique framework--dependent optics.
Starting from two indexed categories, which encode what maps are allowed in the
forward and backward directions, we define a category of dependent optics and
establish under what assumptions it has coproducts. Different choices of
indexed categories correspond to different families of optics: we discuss
dependent lenses and prisms, as well as closed dependent optics. We introduce
the notion of Tambara representation and use it to classify contravariant
functors from the category of optics, thus generalizing the profunctor encoding
of optics to the dependent case.Comment: 17 pages. Submitted to Applied Category Theory 202
Machines of Finite Depth: Towards a Formalization of Neural Networks
We provide a unifying framework where artificial neural networks and their architectures can be formally described as particular cases of a general mathematical construction---machines of finite depth. Unlike neural networks, machines have a precise definition, from which several properties follow naturally. Machines of finite depth are modular (they can be combined), efficiently computable, and differentiable. The backward pass of a machine is again a machine and can be computed without overhead using the same procedure as the forward pass. We prove this statement theoretically and practically via a unified implementation that generalizes several classical architectures---dense, convolutional, and recurrent neural networks with a rich shortcut structure---and their respective backpropagation rules
Generalized Persistence for Equivariant Operators in Machine Learning
Artificial neural networks can learn complex, salient data features to achieve a given task. On the opposite end of the spectrum, mathematically grounded methods such as topological data analysis allow users to design analysis pipelines fully aware of data constraints and symmetries. We introduce an original class of neural network layers based on a generalization of topological persistence. The proposed persistence-based layers allow the users to encode specific data properties (e.g., equivariance) easily. Additionally, these layers can be trained through standard optimization procedures (backpropagation) and composed with classical layers. We test the performance of generalized persistence-based layers as pooling operators in convolutional neural networks for image classification on the MNIST, Fashion-MNIST and CIFAR-10 datasets
Learning to represent signals spike by spike
International audienceNetworks based on coordinated spike coding can encode information with high efficiency in the spike trains of individual neurons. These networks exhibit single-neuron variability and tuning curves as typically observed in cortex, but paradoxically coincide with a precise, non-redundant spike-based population code. However, it has remained unclear whether the specific synaptic connectivities required in these networks can be learnt with local learning rules. Here, we show how to learn the required architecture. Using coding efficiency as an objective, we derive spike-timing-dependent learning rules for a recurrent neural network, and we provide exact solutions for the networks' convergence to an optimal state. As a result, we deduce an entire network from its input distribution and a firing cost. After learning, basic biophysical quantities such as voltages, firing thresholds, excitation, inhibition, or spikes acquire precise functional interpretations